Rename for_each_cpu() to for_each_possible_cpu()
authorKeir Fraser <keir.fraser@citrix.com>
Wed, 15 Jul 2009 12:15:50 +0000 (13:15 +0100)
committerKeir Fraser <keir.fraser@citrix.com>
Wed, 15 Jul 2009 12:15:50 +0000 (13:15 +0100)
... to be more precise in naming, and also to match Linux.

Signed-off-by: Jan Beulich <jbeulich@novell.com>
14 files changed:
xen/arch/ia64/linux-xen/perfmon.c
xen/arch/ia64/linux-xen/smpboot.c
xen/arch/ia64/linux-xen/sn/kernel/sn2_smp.c
xen/arch/ia64/xen/privop_stat.c
xen/arch/ia64/xen/xensetup.c
xen/arch/x86/hpet.c
xen/arch/x86/nmi.c
xen/arch/x86/smpboot.c
xen/arch/x86/time.c
xen/common/perfc.c
xen/common/schedule.c
xen/common/timer.c
xen/common/tmem_xen.c
xen/include/xen/cpumask.h

index 6aed1614087c44568881cbdf75a3b46e9b506c0d..9d31fe91f2d28b6a2bcef2dca607544b1db0758e 100644 (file)
@@ -7313,7 +7313,7 @@ xenpfm_context_create(XEN_GUEST_HANDLE(pfarg_context_t) req)
                goto out;
 
        /* XXX fmt */
-       for_each_cpu(cpu) {
+       for_each_possible_cpu(cpu) {
                ctx[cpu] = pfm_context_create(&kreq);
                if (ctx[cpu] == NULL) {
                        error = -ENOMEM;
@@ -7325,20 +7325,20 @@ xenpfm_context_create(XEN_GUEST_HANDLE(pfarg_context_t) req)
 
        BUG_ON(in_irq());
        spin_lock(&xenpfm_context_lock);
-       for_each_cpu(cpu) {
+       for_each_possible_cpu(cpu) {
                if (per_cpu(xenpfm_context, cpu) != NULL) {
                        error = -EBUSY;
                        break;
                }
        }
-       for_each_cpu(cpu) {
+       for_each_possible_cpu(cpu) {
                per_cpu(xenpfm_context, cpu) = ctx[cpu];
                ctx[cpu] = NULL;
        }
        spin_unlock(&xenpfm_context_lock);
 
 out:
-       for_each_cpu(cpu) {
+       for_each_possible_cpu(cpu) {
                if (ctx[cpu] != NULL)
                        pfm_context_free(ctx[cpu]);
        }
@@ -7358,7 +7358,7 @@ again:
        need_unload = 0;
        BUG_ON(in_irq());
        spin_lock_irqsave(&xenpfm_context_lock, flags);
-       for_each_cpu(cpu) {
+       for_each_possible_cpu(cpu) {
                ctx = per_cpu(xenpfm_context, cpu);
                if (ctx == NULL) {
                        error = -EINVAL;
@@ -7369,7 +7369,7 @@ again:
                        need_unload = 1;
        }
        if (error) {
-               for_each_cpu(cpu) {
+               for_each_possible_cpu(cpu) {
                        ctx = per_cpu(xenpfm_context, cpu);
                        if (ctx == NULL)
                                break;
@@ -7378,7 +7378,7 @@ again:
                goto out;
        }
        if (need_unload) {
-               for_each_cpu(cpu)
+               for_each_possible_cpu(cpu)
                        UNPROTECT_CTX_NOIRQ(per_cpu(xenpfm_context, cpu));
                spin_unlock_irqrestore(&xenpfm_context_lock, flags);
 
@@ -7388,7 +7388,7 @@ again:
                goto again;
        }
 
-       for_each_cpu(cpu) {
+       for_each_possible_cpu(cpu) {
                pfm_context_t* ctx = per_cpu(xenpfm_context, cpu);
                per_cpu(xenpfm_context, cpu) = NULL;
 
@@ -7740,7 +7740,7 @@ xenpfm_start_stop_locked(int is_start)
        arg.is_start = is_start;
        atomic_set(&arg.started, 1); /* 1 for this cpu */
        atomic_set(&arg.finished, 0);
-       for_each_cpu(cpu)
+       for_each_possible_cpu(cpu)
                arg.error[cpu] = 0;
 
        BUG_ON(!spin_is_locked(&xenpfm_context_lock));
index f9cd243b3f635b07a1a7c1f2d8f98d9ca831d64e..805706241e9f127514c7de81a158707d94928339 100644 (file)
@@ -675,7 +675,7 @@ remove_from_mtinfo(int cpu)
 {
        int i;
 
-       for_each_cpu(i)
+       for_each_possible_cpu(i)
                if (mt_info[i].valid &&  mt_info[i].socket_id ==
                                                cpu_data(cpu)->socket_id)
                        mt_info[i].valid = 0;
@@ -874,7 +874,7 @@ check_for_mtinfo_index(void)
 {
        int i;
        
-       for_each_cpu(i)
+       for_each_possible_cpu(i)
                if (!mt_info[i].valid)
                        return i;
 
@@ -892,7 +892,7 @@ check_for_new_socket(__u16 logical_address, struct cpuinfo_ia64 *c)
        int i;
        __u32 sid = c->socket_id;
 
-       for_each_cpu(i) {
+       for_each_possible_cpu(i) {
                if (mt_info[i].valid && mt_info[i].proc_fixed_addr == logical_address
                    && mt_info[i].socket_id == sid) {
                        c->core_id = mt_info[i].core_id;
index 4626b57187f9366d6dcfc86ab9cf7de50d96fad5..aae029c7b22310bb6359b4b8b32de52832c2e767 100644 (file)
@@ -211,7 +211,7 @@ sn2_global_tlb_purge(unsigned long start,
        spin_lock(&sn2_ptcg_lock2);
        node_set(cpu_to_node(smp_processor_id()), nodes_flushed);
        i = 0;
-       for_each_cpu(cpu) {
+       for_each_possible_cpu(cpu) {
                cnode = cpu_to_node(cpu);
                if (!node_isset(cnode, nodes_flushed)) {
                        cpu_set(cpu, selected_cpus);
@@ -269,7 +269,7 @@ sn2_global_tlb_purge(struct mm_struct *mm, unsigned long start,
                i++;
        }
 #else
-       for_each_cpu(cpu) {
+       for_each_possible_cpu(cpu) {
                cnode = cpu_to_node(cpu);
                node_set(cnode, nodes_flushed);
                lcpu = cpu;
index bfe81aaff7887e466cb98425195d76ed7c35777a..e8e6fa992152c3cacea1e63cda67e1e337d27a12 100644 (file)
@@ -65,7 +65,7 @@ void gather_privop_addrs(void)
 {
        unsigned int cpu;
 
-       for_each_cpu ( cpu ) {
+       for_each_possible_cpu ( cpu ) {
                perfc_t *perfcounters = per_cpu(perfcounters, cpu);
                struct privop_addr_count *s = per_cpu(privop_addr_counter, cpu);
                int i, j;
@@ -92,7 +92,7 @@ void reset_privop_addrs(void)
 {
        unsigned int cpu;
 
-       for_each_cpu ( cpu ) {
+       for_each_possible_cpu ( cpu ) {
                struct privop_addr_count *v = per_cpu(privop_addr_counter, cpu);
                int i, j;
 
index bde96cab0e4fe827fe6f7992e397e3f3834a359c..579fe3c86129530e6825287d4c72afa69347e594 100644 (file)
@@ -606,8 +606,7 @@ skip_move:
     smp_prepare_cpus(max_cpus);
 
     /* We aren't hotplug-capable yet. */
-    for_each_cpu ( i )
-        cpu_set(i, cpu_present_map);
+    cpus_or(cpu_present_map, cpu_present_map, cpu_possible_map);
 
     /*  Enable IRQ to receive IPI (needed for ITC sync).  */
     local_irq_enable();
index 6877a80aead435cbb4a8ba5e5a2525ff33f53c4c..5ea752348baebbd18b27398dcb5dd7a52225813b 100644 (file)
@@ -590,7 +590,7 @@ void hpet_broadcast_init(void)
     legacy_hpet_event.flags = 0;
     spin_lock_init(&legacy_hpet_event.lock);
 
-    for_each_cpu(i)
+    for_each_possible_cpu(i)
         per_cpu(cpu_bc_channel, i) = &legacy_hpet_event;
 
     if ( !force_hpet_broadcast )
index 9db89999651b6a0f8fc53eb1fe8f5d34fc124b34..704fb8ee9853f73bc9c40bb360c6a8da6798523b 100644 (file)
@@ -460,7 +460,7 @@ static void do_nmi_stats(unsigned char key)
     struct vcpu *v;
 
     printk("CPU\tNMI\n");
-    for_each_cpu ( i )
+    for_each_possible_cpu ( i )
         printk("%3d\t%3d\n", i, nmi_count(i));
 
     if ( ((d = dom0) == NULL) || (d->vcpu == NULL) ||
index 203c3682f67f1aa935d68012140e4f5f285e4d8d..420c0864269dc79239619de7eb0831ff221d9c68 100644 (file)
@@ -1163,7 +1163,7 @@ static void __init smp_boot_cpus(unsigned int max_cpus)
         * construct cpu_sibling_map, so that we can tell sibling CPUs
         * efficiently.
         */
-       for_each_cpu(cpu) {
+       for_each_possible_cpu(cpu) {
                cpus_clear(per_cpu(cpu_sibling_map, cpu));
                cpus_clear(per_cpu(cpu_core_map, cpu));
        }
index 48d4ee1e9d7d1495a6608ae5fd10db3ddf0d42db..8fc2a4836b16cd2685c10975bf165c20a2406e70 100644 (file)
@@ -1227,7 +1227,7 @@ int __init init_xen_time(void)
     if ( opt_consistent_tscs )
     {
         int cpu;
-        for_each_cpu ( cpu )
+        for_each_possible_cpu ( cpu )
             per_cpu(cpu_time, cpu).tsc_scale = per_cpu(cpu_time, 0).tsc_scale;
     }
 
index fe0958b30c28432b9fd82db123f64439c9889488..5e1e196a5f0c1686c329d0b5aa5ee04a9e4979cf 100644 (file)
@@ -136,13 +136,13 @@ void perfc_reset(unsigned char key)
         switch ( perfc_info[i].type )
         {
         case TYPE_SINGLE:
-            for_each_cpu ( cpu )
+            for_each_possible_cpu ( cpu )
                 per_cpu(perfcounters, cpu)[j] = 0;
         case TYPE_S_SINGLE:
             ++j;
             break;
         case TYPE_ARRAY:
-            for_each_cpu ( cpu )
+            for_each_possible_cpu ( cpu )
                 memset(per_cpu(perfcounters, cpu) + j, 0,
                        perfc_info[i].nr_elements * sizeof(perfc_t));
         case TYPE_S_ARRAY:
@@ -205,14 +205,14 @@ static int perfc_copy_info(XEN_GUEST_HANDLE_64(xen_sysctl_perfc_desc_t) desc,
         {
         case TYPE_SINGLE:
         case TYPE_S_SINGLE:
-            for_each_cpu ( cpu )
+            for_each_possible_cpu ( cpu )
                 perfc_vals[v++] = per_cpu(perfcounters, cpu)[j];
             ++j;
             break;
         case TYPE_ARRAY:
         case TYPE_S_ARRAY:
             memset(perfc_vals + v, 0, perfc_d[i].nr_vals * sizeof(*perfc_vals));
-            for_each_cpu ( cpu )
+            for_each_possible_cpu ( cpu )
             {
                 perfc_t *counters = per_cpu(perfcounters, cpu) + j;
                 unsigned int k;
index fb2464f11d53060a2ee8824d80c87c2e47c7ddbb..9b59ea6d8423f20e7bb02bed34c43811116d045e 100644 (file)
@@ -918,7 +918,7 @@ void __init scheduler_init(void)
 
     open_softirq(SCHEDULE_SOFTIRQ, schedule);
 
-    for_each_cpu ( i )
+    for_each_possible_cpu ( i )
     {
         spin_lock_init(&per_cpu(schedule_data, i).schedule_lock);
         init_timer(&per_cpu(schedule_data, i).s_timer, s_timer_fn, NULL, i);
index eca25b3fb75664f0778daf7987fbf1dc97fa1509..8f975958d8859b8f152bc8d23ee8cfa87b036940 100644 (file)
@@ -529,7 +529,7 @@ void __init timer_init(void)
     SET_HEAP_SIZE(&dummy_heap, 0);
     SET_HEAP_LIMIT(&dummy_heap, 0);
 
-    for_each_cpu ( i )
+    for_each_possible_cpu ( i )
     {
         spin_lock_init(&per_cpu(timers, i).lock);
         per_cpu(timers, i).heap = &dummy_heap;
index f1b94239f4de1924522a8670cea797738318a82d..f813b0e52fa69e12bef2ae7956347efee0a2f209 100644 (file)
@@ -317,7 +317,7 @@ EXPORT int tmh_init(void)
 
     dstmem_order = get_order_from_pages(LZO_DSTMEM_PAGES);
     workmem_order = get_order_from_bytes(LZO1X_1_MEM_COMPRESS);
-    for_each_cpu ( cpu )
+    for_each_possible_cpu ( cpu )
     {
         pi = alloc_domheap_pages(0,dstmem_order,0);
         per_cpu(dstmem, cpu) = p1 = ((pi == NULL) ? NULL : page_to_virt(pi));
index 43f538eba44295e16a16feeac0923ce3286e25ab..40222fd6cd271ca00d834ff51ed0421a4d9d0591 100644 (file)
@@ -61,7 +61,7 @@
  *
  * int any_online_cpu(mask)            First online cpu in mask, or NR_CPUS
  *
- * for_each_cpu(cpu)                   for-loop cpu over cpu_possible_map
+ * for_each_possible_cpu(cpu)          for-loop cpu over cpu_possible_map
  * for_each_online_cpu(cpu)            for-loop cpu over cpu_online_map
  * for_each_present_cpu(cpu)           for-loop cpu over cpu_present_map
  *
@@ -402,9 +402,9 @@ extern cpumask_t cpu_present_map;
        cpu;                                    \
 })
 
-#define for_each_cpu(cpu)        for_each_cpu_mask((cpu), cpu_possible_map)
-#define for_each_online_cpu(cpu)  for_each_cpu_mask((cpu), cpu_online_map)
-#define for_each_present_cpu(cpu) for_each_cpu_mask((cpu), cpu_present_map)
+#define for_each_possible_cpu(cpu) for_each_cpu_mask((cpu), cpu_possible_map)
+#define for_each_online_cpu(cpu)   for_each_cpu_mask((cpu), cpu_online_map)
+#define for_each_present_cpu(cpu)  for_each_cpu_mask((cpu), cpu_present_map)
 
 /* Copy to/from cpumap provided by control tools. */
 struct xenctl_cpumap;